# Chain-of-Thought Reasoning

Mmada 8B MixCoT
MIT
MMaDA is a novel class of multimodal diffusion foundation models, excelling in various domains such as text reasoning, multimodal understanding, and text-to-image generation.
Text-to-Image Transformers
M
Gen-Verse
601
3
Deepseek R1 Llama 8B F32 GGUF
Apache-2.0
DeepSeek-R1-Llama-8B-F32-GGUF is the quantized version of DeepSeek-R1-Distill-Llama-8B, trained directly with reinforcement learning, featuring capabilities such as self-verification, reflection, and generating extended chain-of-thought reasoning.
Large Language Model Transformers English
D
prithivMLmods
326
1
Sarvamai Sarvam M GGUF
Apache-2.0
This is a quantized version of the Sarvam-m model, supporting text generation tasks in multiple Indian languages and English.
Large Language Model Supports Multiple Languages
S
bartowski
845
1
Qwenstoryteller I1 GGUF
Apache-2.0
QwenStoryteller is a storytelling model based on a vision-language model, focusing on visual narrative and cross-frame consistency.
Image-to-Text English
Q
mradermacher
340
0
Qwenstoryteller GGUF
Apache-2.0
Quantized version of Qwen's visual storytelling model, focusing on cross-frame consistent story generation and image-to-text tasks
Image-to-Text English
Q
mradermacher
195
0
Qwen3 235B A22B AWQ
Apache-2.0
Qwen3-235B-A22B is the latest generation large language model in the Qwen series, adopting a Mixture of Experts (MoE) architecture with 235 billion parameters and 22 billion active parameters. It excels in reasoning, instruction following, agent capabilities, and multilingual support.
Large Language Model Transformers
Q
cognitivecomputations
2,563
9
MAI DS R1 GGUF
MIT
MAI-DS-R1 is the DeepSeek-R1 inference model, further trained by Microsoft's AI team to enhance its responsiveness on restricted topics and optimize its risk performance while maintaining its reasoning capabilities and competitive performance.
Large Language Model
M
unsloth
916
4
Videochat R1 7B Caption
Apache-2.0
VideoChat-R1_7B_caption is a multimodal video-text generation model based on Qwen2-VL-7B-Instruct, focusing on video content understanding and description generation.
Video-to-Text Transformers English
V
OpenGVLab
48
1
UIGEN T1.5 32B
Apache-2.0
UIGEN-T1.5 is an advanced Transformer UI generation model fine-tuned from Qwen2.5-Coder-32B-Instruct, specifically designed for generating modern and unique front-end user interfaces.
Large Language Model Transformers English
U
Tesslate
44
4
UIGEN T1.5 7B
Apache-2.0
An advanced UI generation model fine-tuned based on Qwen2.5-Coder-7B-Instruct, excelling in generating modern and visually appealing HTML/CSS code.
Large Language Model Transformers English
U
Tesslate
121
5
UIGEN T1.5 14B
Apache-2.0
An advanced Transformer-based interface generation model fine-tuned on Qwen2.5-Coder-14B-Instruct, specializing in generating modern and unique frontend user interfaces
Large Language Model Transformers English
U
Tesslate
67
5
Deepseek R1 14b Cot Math Reasoning Full GGUF
MIT
14B-parameter-scale mathematical reasoning large language model supporting chain-of-thought reasoning
Large Language Model
D
tensorblock
838
2
Deephermes 3 Llama 3 3B Preview Abliterated
Apache-2.0
DeepHermes 3 Preview Abliterated is the latest version of Nous Research's flagship Hermes series large language model, integrating reasoning and conventional LLM response modes.
Large Language Model Transformers English
D
prithivMLmods
338
2
Gemma 3 4b Reasoning
Apache-2.0
Gemma-3-4b Reasoning is a Transformer-based language model fine-tuned using the GRPO method, specializing in reasoning task optimization.
Large Language Model Transformers English
G
ericrisco
53
2
Traceback 12b
Apache-2.0
TraceBack 12b is a 4bit quantized version based on the Mistral-Nemo-Instruct architecture, focusing on instruction-following and chain-of-thought reasoning tasks.
Large Language Model Transformers
T
secemp9
1,470
29
Tifa DeepsexV2 7b MGRPO Safetensors GGUF
Apache-2.0
Tifa-DeepsexV2-7b-MGRPO-safetensors is a multilingual (Chinese and English) large language model based on the transformers library, optimized through incremental pre-training, supervised fine-tuning, and reinforcement learning, suitable for role-playing and chain-of-thought tasks.
Large Language Model Supports Multiple Languages
T
mradermacher
283
1
Minithinky V2 1B Llama 3.2
MiniThinky 1B is an AI assistant model fine-tuned from Llama-3.2-1B-Instruct, featuring thought process visualization before answering
Large Language Model Transformers
M
ngxson
260
40
MD Judge V0 2 Internlm2 7b
Apache-2.0
A safety protection tool fine-tuned based on internlm2-7b-chat, providing human-readable judgment explanations and fine-grained harm scoring
Large Language Model Transformers English
M
OpenSafetyLab
1,823
15
Flan T5 Tsa Thor Xl
MIT
A targeted sentiment analysis model fine-tuned from Flan-T5-XL, trained using the Three-Hop Reasoning (THoR) framework, specifically designed for English text
Large Language Model Transformers English
F
nicolay-r
34
3
Llama 3 Cat 8b Instruct V1
A fine-tuned model based on llama 3 8b, focusing on system prompt fidelity, helpfulness, and role immersion
Large Language Model Transformers
L
SteelStorage
20
51
Quietstar 8 Ahead
Based on the Mistral-7b model, it employs the Quiet-STaR method for continuous pretraining, generating 8 reasoning tokens before each output token to enhance reasoning capabilities.
Large Language Model Transformers
Q
ezelikman
239
90
Nli Entailment Verifier Xxl
An NLI model fine-tuned based on flan-t5-xxl, used to verify whether a premise supports a hypothesis, specially optimized for multi-sentence premise scenarios
Large Language Model Transformers English
N
soumyasanyal
164
5
Synthia 70B V1.5
Synthia-70B-v1.5 is a 70-billion-parameter large language model based on the Llama2 architecture, focusing on complex reasoning and coherent responses through the Tree of Thought method.
Large Language Model Transformers
S
migtissera
99
42
Llemma 34b
Llemma 34B is a language model specialized in the field of mathematics, initialized based on the weights of Code Llama 34B and trained on 50 billion tokens from the Proof-Pile-2 dataset.
Large Language Model Transformers English
L
EleutherAI
60
99
Synthia 70B V1.2b
SynthIA (Synthetic Intelligence Agent) is an LLama-2-70B model trained on an Orca-style dataset, excelling in instruction following and long dialogues.
Large Language Model Transformers English
S
migtissera
136
29
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase